Variational Autoencoder

A generative model that learns a latent representation by combining an encoder and a decoder . Training maximizes the ELBO (Evidence Lower Bound):

The first term is reconstruction quality; the KL term regularizes the latent space toward the prior (typically a standard Gaussian). In the rate-distortion view, the relevant rate term is the data-averaged KL, .

Variations

-VAE

Weights the KL term by :

tightens the latent bottleneck and typically increases disentanglement pressure, often at the cost of reconstruction quality. Disentangling means latent dimensions align more with independent factors of variation (e.g., color, shape, size) rather than mixing them.

This directly controls the Rate Distortion theory tradeoff: sets the compression-fidelity balance.